A Laplacian approach to $$\ell _1$$-norm minimization
نویسندگان
چکیده
We propose a novel differentiable reformulation of the linearly-constrained $\ell_1$ minimization problem, also known as basis pursuit problem. The is inspired by Laplacian paradigm network theory and leads to new family gradient-based methods for solution problems. analyze iteration complexity natural approach reformulation, based on multiplicative weights update scheme, well an accelerated gradient scheme. results can be seen bounds iteratively reweighted least squares (IRLS) type pursuit.
منابع مشابه
Mixed l0/l1 Norm Minimization Approach to Super-Resolution
This deals with the problem of recovering a high-resolution digital image from one low resolution digital image and proposes a super-resolution algorithm based on the mixed l0/l1 norm minimization. Introducing some assumptions and focusing the uniformity and the gradation of the image, this paper formulates the colorization problem as a mixed l0/l1 norm minimization and proposes the algorithm b...
متن کاملPerformance of first- and second-order methods for \(\ell _1\) -regularized least squares problems
We study the performance of firstand second-order optimization methods for `1-regularized sparse least-squares problems as the conditioning of the problem changes and the dimensions of the problem increase up to one trillion. A rigorously defined generator is presented which allows control of the dimensions, the conditioning and the sparsity of the problem. The generator has very low memory req...
متن کاملMaximum Consensus Parameter Estimation by Reweighted \ell _1 ℓ 1 Methods
Robust parameter estimation in computer vision is frequently accomplished by solving the maximum consensus (MaxCon) problem. Widely used randomized methods for MaxCon, however, can only produce random approximate solutions, while global methods are too slow to exercise on realistic problem sizes. Here we analyse MaxCon as iterative reweighted algorithms on the data residuals. We propose a smoot...
متن کاملRecovery of sparsest signals via $\ell^q$-minimization
In this paper, it is proved that every s-sparse vector x ∈ R can be exactly recovered from the measurement vector z = Ax ∈ R via some l-minimization with 0 < q ≤ 1, as soon as each s-sparse vector x ∈ R n is uniquely determined by the measurement z.
متن کاملA max-norm constrained minimization approach to 1-bit matrix completion
We consider in this paper the problem of noisy 1-bit matrix completion under a general non-uniform sampling distribution using the max-norm as a convex relaxation for the rank. A max-norm constrained maximum likelihood estimate is introduced and studied. The rate of convergence for the estimate is obtained. Information-theoretical methods are used to establish a minimax lower bound under the ge...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
ژورنال
عنوان ژورنال: Computational Optimization and Applications
سال: 2021
ISSN: ['0926-6003', '1573-2894']
DOI: https://doi.org/10.1007/s10589-021-00270-x